Current:Home > MarketsCalifornia advances measures targeting AI discrimination and deepfakes -Capitatum
California advances measures targeting AI discrimination and deepfakes
NovaQuant Quantitative Think Tank Center View
Date:2025-04-06 07:25:43
SACRAMENTO, Calif. (AP) — As corporations increasingly weave artificial intelligence technologies into the daily lives of Americans, California lawmakers want to build public trust, fight algorithmic discrimination and outlaw deepfakes that involve elections or pornography.
The efforts in California — home to many of the world’s biggest AI companies — could pave the way for AI regulations across the country. The United States is already behind Europe in regulating AI to limit risks, lawmakers and experts say, and the rapidly growing technology is raising concerns about job loss, misinformation, invasions of privacy and automation bias.
A slew of proposals aimed at addressing those concerns advanced last week, but must win the other chamber’s approval before arriving at Gov. Gavin Newsom’s desk. The Democratic governor has promoted California as an early adopter as well as regulator, saying the state could soon deploy generative AI tools to address highway congestion, make roads safer and provide tax guidance, even as his administration considers new rules against AI discrimination in hiring practices.
With strong privacy laws already in place, California is in a better position to enact impactful regulations than other states with large AI interests, such as New York, said Tatiana Rice, deputy director of the Future of Privacy Forum, a nonprofit that works with lawmakers on technology and privacy proposals.
“You need a data privacy law to be able to pass an AI law,” Rice said. “We’re still kind of paying attention to what New York is doing, but I would put more bets on California.”
California lawmakers said they cannot wait to act, citing hard lessons they learned from failing to reign in social media companies when they might have had a chance. But they also want to continue attracting AI companies to the state.
Here’s a closer look at California’s proposals:
FIGHTING AI DISCRIMINATION AND BUILDING PUBLIC TRUST
Some companies, including hospitals, already use AI models to define decisions about hiring, housing and medical options for millions of Americans without much oversight. Up to 83% of employers are using AI to help in hiring, according to the U.S. Equal Employment Opportunity Commission. How those algorithms work largely remains a mystery.
One of the most ambitious AI measures in California this year would pull back the curtains on these models by establishing an oversight framework to prevent bias and discrimination. It would require companies using AI tools to participate in decisions that determine results and to inform people affected when AI is used. AI developers would have to routinely make internal assessments of their models for bias. And the state attorney general would have authority to investigate reports of discriminating models and impose fines of $10,000 per violation.
AI companies also might soon be required to start disclosing what data they’re using to train their models.
PROTECTING JOBS AND LIKENESS
Inspired by the months-long Hollywood actors strike last year, a California lawmaker wants to protect workers from being replaced by their AI-generated clones — a major point of contention in contract negotiations.
The proposal, backed by the California Labor Federation, would let performers back out of existing contracts if vague language might allow studios to freely use AI to digitally clone their voices and likeness. It would also require that performers be represented by an attorney or union representative when signing new “voice and likeness” contracts.
California may also create penalties for digitally cloning dead people without the consent of their estate, citing the case of a media company that produced a fake, AI-generated hourlong comedy special to recreate the late comedian George Carlin’s style and material without his estate’s permission.
REGULATING POWERFUL GENERATIVE AI SYSTEMS
Real-world risks abound as generative AI creates new content such as text, audio and photos in response to prompts. So lawmakers are considering requiring guardrails around “extremely large” AI systems that have the potential to spit out instructions for creating disasters — such as building chemical weapons or assisting in cyberattacks — that could cause at least $500 million in damages. It would require such models to have a built-in “kill switch,” among other things.
The measure, supported by some of the most renowned AI researchers, would also create a new state agency to oversee developers and provide best practices, including for still-more powerful models that don’t yet exist. The state attorney general also would be able to pursue legal actions in case of violations.
BANNING DEEPFAKES INVOLVING POLITICS OR PORNOGRAPHY
A bipartisan coalition seeks to facilitate prosecuting people who use AI tools to create images of child sexual abuse. Current law does not allow district attorneys to go after people who possess or distribute AI-generated child sexual abuse images if the materials are not depicting a real person, law enforcement said.
A host of Democratic lawmakers are also backing a bill tackling election deepfakes, citing concerns after AI-generated robocalls mimicked President Joe Biden’s voice ahead of New Hampshire’s recent presidential primary. The proposal would ban “materially deceptive” deepfakes related to elections in political mailers, robocalls and TV ads 120 days before Election Day and 60 days thereafter. Another proposal would require social media platforms to label any election-related posts created by AI.
veryGood! (8)
Related
- NCAA hits former Michigan coach Jim Harbaugh with suspension, show-cause for recruiting violations
- Peru lost more than half of its glacier surface in just over half a century, scientists say
- Barclay Briggs, backup FCS lineman, finds following with hilarious NFL draft declaration
- Georgia Supreme Court ruling prevents GOP-backed commission from beginning to discipline prosecutors
- Mega Millions winning numbers for August 6 drawing: Jackpot climbs to $398 million
- Salty much? These brain cells decide when tasty becomes blech
- Former Penthouse magazine model sues Axl Rose of Guns N’ Roses, saying he raped her in 1989
- Watch man travel 1200 miles to reunite with long-lost dog after months apart
- Man can't find second winning lottery ticket, sues over $394 million jackpot, lawsuit says
- Colts LB Shaquille Leonard stunned by release, still shows up for turkey drive
Ranking
- Immigration issues sorted, Guatemala runner Luis Grijalva can now focus solely on sports
- Utah Tech women’s hoops coach suspended for 2 games after investigation based on player complaints
- 13 Secrets About Mrs. Doubtfire Are on the Way, Dear
- A crane operator has rescued a man from a burning high-rise in England
- Everything Simone Biles did at the Paris Olympics was amplified. She thrived in the spotlight
- French military to contribute 15,000 soldiers to massive security operation for Paris Olympics
- What’s That on Top of the Building? A New Solar Water Heating System Goes Online as Its Developer Enters the US Market
- AP Week in Pictures: Latin America and Caribbean
Recommendation
Krispy Kreme offers a free dozen Grinch green doughnuts: When to get the deal
AP Week in Pictures: Global | Nov. 17 - Nov. 23, 2023
Zach Edey's MVP performance leads No. 2 Purdue to Maui Invitational title
What Happened to the Great Lakes Offshore Wind Boom?
British golfer Charley Hull blames injury, not lack of cigarettes, for poor Olympic start
4-day truce begins in Israel-Hamas war, sets stage for release of dozens of Gaza-held hostages
How the hostage deal came about: Negotiations stumbled, but persistence finally won out
New Mexico Supreme Court reprimands judge who advised prosecutors in case involving his daughter